Web Survey Bibliography
Background
The number of online surveys in German official statistics has been steadily growing over the years. So far, the CAWI-mode is mainly implemented in business and administrative statistics, but there is also a huge demand for online questionnaires in social statistics. Reasons are manifold: E.g. restrictions of financing and time, also expectations of the population to respond to questionnaires online, play an important role.
Hence, to guarantee high quality of statistical data, online questionnaires need to be easy to handle for respondents. Consequently a sophisticated and thoughtful programming is needed. However, only building functionality into a product doesn’t guarantee that people will be able to use it. Therefore the Federal Statistical Office implements usability-tests with online questionnaires to include the view of respondents into consideration.
Methodology
In usability-tests in the Federal Statistical Office a three step approach is applied: Firstly, we observe eye movements and facial expressions, while respondents deal with the questionnaire. Secondly, we conduct cognitive interviews, in which we show the video of eye movements during the fill-in process to the probands. By using cognitive methods we try to figure out reasons for incomplete or missing answers. In a third step, we analyse and interpret the eye tracking data. As there are benefits from every method on the one hand, each method has its weaknesses, too. Therefore we combine information from the three sources with each other (“triangulation”), to provide higher data quality in terms of more objective results and a richer overall picture.
Results
Based on the results of four usability-tests and the findings from current research, the Federal Statistical Office develops a profile of requirements for online questionnaire design in social statistics. The ongoing project addresses items such as the general format of the questionnaire, to design und functionality of navigation, explanations, completion instructions, error messages, question and answering formats and accessibility.
The presentation describes the process of pretesting online-questionnaires and the ‘(re)-search of basic requirements’. Background of applying this term is that, when defining a profile of requirements, it is linked to the basic experience, that even thoroughly selected design approaches by researchers may not meet users’ needs in practice. Through the eye of the respondent, pretests deliver insights into survey reality and therefore complement results from research literature. Taking results from pretests together with facts from research, lead us to conclusions on how to enhance the design of online-questionnaires.
Conference Homepage (abstract) / (full text)
Web survey bibliography (4086)
- Media tracker; 2012
- Measuring the quality of governmental websites in a controlled versus an online setting with the ‘...; 2012; Elling, S., Lentz, L., de Jong , M., van den Bergh, H.
- Measuring modern media consumption; 2012; Arini, N.
- ISO 20252. Market, opinion and social research-Vocabulary and service requirements, 2nd Edition; 2012
- Is „chapterisation“ a viable alternative to traditional progress indicators ?; 2012; Spicer, R., Dowling, Z.
- Internet use in households and by individual in 2012. Eurostat Statistics in Focus 50/2012; 2012; Seybert, H.
- Internet access - Households and individuals, 2012 part 2; 2012
- Internet access - Households and individuals, 2012; 2012
- Google et Médiamétrie créent une audience bimédia; 2012; Gonzales, P.
- GMI Pinnacle; 2012
- Global market research 2012; 2012
- Explaining rising nonresponse rates in cross-sectional surveys; 2012; Brick, J. M., Williams, Do.
- Eurobarometer Special surveys: Special Eurobarometer 381; 2012
- Online Surveys 2.0; 2012; Elferink, R.
- The Impact of Academic Sponsorship on Online Survey Dropout Rates; 2012; Allen, P. J., Roberts, L. D.
- Especially for You: Motivating Respondents in an Internet Panel by Offering Tailored Questions; 2012; Oudejans, M.
- Social media as a data collection tool: the impact of Facebook in behavioural research; 2012; Zoppos, E.
- Smartphone Apps and User Engagement: Collecting Data in the Digital Era; 2012; Link, M. W.
- Snowball Sampling in Online Social Networks; 2012; Raissi, M., Ackland, R.
- The Use of Facebook as a Locating and Contacting Tool; 2012; McCarthy, T.
- How Often Do You Use the App with a Bird on It? Exploring Differences in Survey Completion Times, Primacy...; 2012; Buskirk, T. D.
- Data quality of questions sensitive to social-desirability bias in web surveys; 2012; Lozar Manfreda, K., Zajc, N., Berzelak, N., Vehovar, V.
- Online Questionnaires: Development of ‘basic requirements’; 2012; Tries, S., Blanke, K.
- Social research in online context: methodological reflections on web surveys from a case study; 2012; Pandolfini, V.
- Efficacy of a health-related Facebook social network site on health-seeking behaviors; 2012; Woolley, P., Peterson, M.
- The war against unengaged online respondents; 2012; Gittelman, S. H., Trimarchi, E.
- Qualitatively Speaking: The five absolute, no-excuse must-dos for online qualitative researchers; 2012; Rossow, A.
- By the Numbers: Lessons for using online panels in B2B research; 2012; Elsner, N.
- Specialized Tools for Measuring Past Events ; 2012; Belli, R. F.
- Transparency, Access and the Credibility of Survey Research; 2012; Lupia, A.
- Can Microtargeting Improve Survey Sampling? An Assessment of Accuracy and Bias in Consumer File Marketing...; 2012; Pasek, J.
- Anonymity and Confidentiality; 2012; Tourangeau, R.
- Cognitive Evaluation of Survey Instruments: State of the Science (Art?) and Future Directions; 2012; Willis, G. B.
- Oh, Just One More Thing … Leveraging “Leave-Behinds” in Data Collection; 2012; Link, M. W.
- Paradata; 2012; Kreuter, F.
- Computation of Survey Weights: Bridging Theory and Practice; 2012; DeBell, M.
- Optimizing Response Rates; 2012; Brick, J. M.
- Modes of Data Collection; 2012; Tourangeau, R.
- The Use and Effects of Incentives in Surveys; 2012; Singer, E.
- Improving Question Design to Maximize Reliability and Validity; 2012; Krosnick, J. A.
- Respondent Attrition vs Data Attrition and Their Reduction; 2012; Olsen, R. J.
- Survey Interviewing: Deviations from the Script; 2012; Schaeffer, N. C.
- How accurate are surveys of objective phenomena?; 2012; Chang, L. C., Krosnick, J. A.
- Measure the response burden in the Swedish Intrastat system; 2012; Weideskog, F.
- Mode and non-response effects and their treatment; 2012; Chrysanthopoulos, S., Georgostathi, A.
- What can be said about quality in the Central Population Register based on a self-completion survey...; 2012; Falnes-Dalheim, E., Pedersen, H. E.
- Improving the quality of complex surveys: The case of the EU Labour Force Survey ; 2012; van der Valk, J.
- Pros and cons of Internet based User Satisfaction Surveys; 2012; Consoli, A., Matsulevits, L.
- Between demand and reality: Ensuring efficiency and quality in pretesting questionnaires; 2012; Sattelberger, S., Blanke, K.
- How to provide high data quality in online-questionnaires: Setting guidelines in design; 2012; Tries, S., Nebel, S., Blanke, K.